dialogue example
Stephanie: Step-by-Step Dialogues for Mimicking Human Interactions in Social Conversations
Yang, Hao, Lu, Hongyuan, Zeng, Xinhua, Liu, Yang, Zhang, Xiang, Yang, Haoran, Zhang, Yumeng, Huang, Shan, Wei, Yiran, Lam, Wai
In the rapidly evolving field of natural language processing, dialogue systems primarily employ a single-step dialogue paradigm. Although this paradigm is efficient, it lacks the depth and fluidity of human interactions and does not appear natural. We introduce a novel \textbf{Step}-by-Step Dialogue Paradigm (Stephanie), designed to mimic the ongoing dynamic nature of human conversations. By employing a dual learning strategy and a further-split post-editing method, we generated and utilized a high-quality step-by-step dialogue dataset to fine-tune existing large language models, enabling them to perform step-by-step dialogues. We thoroughly present Stephanie. Tailored automatic and human evaluations are conducted to assess its effectiveness compared to the traditional single-step dialogue paradigm. We will release code, Stephanie datasets, and Stephanie LLMs to facilitate the future of chatbot eras.
- North America > Canada > Ontario > Toronto (0.14)
- North America > United States > Washington > King County > Seattle (0.04)
- Europe > Middle East > Malta > Eastern Region > Northern Harbour District > St. Julian's (0.04)
- (2 more...)
Refine and Imitate: Reducing Repetition and Inconsistency in Persuasion Dialogues via Reinforcement Learning and Human Demonstration
Shi, Weiyan, Li, Yu, Sahay, Saurav, Yu, Zhou
Despite the recent success of large-scale language models on various downstream NLP tasks, the repetition and inconsistency problems still persist in dialogue response generation. Previous approaches have attempted to avoid repetition by penalizing the language model's undesirable behaviors in the loss function. However, these methods focus on token-level information and can lead to incoherent responses and uninterpretable behaviors. To alleviate these issues, we propose to apply reinforcement learning to refine an MLE-based language model without user simulators, and distill sentence-level information about repetition, inconsistency and task relevance through rewards. In addition, to better accomplish the dialogue task, the model learns from human demonstration to imitate intellectual activities such as persuasion, and selects the most persuasive responses. Experiments show that our model outperforms previous state-of-the-art dialogue models on both automatic metrics and human evaluation results on a donation persuasion task, and generates more diverse, consistent and persuasive conversations according to the user feedback.
- North America > Puerto Rico (0.04)
- North America > United States > California > Yolo County > Davis (0.04)
Voice assistants – strategies for handling private information
In the latest in this series of posts, researchers from the EU-funded COMPRISE project write about privacy issues associated with voice assistants. They propose possible ways to maintain the privacy of users whilst ensuring that manufacturers can still access the quality usage data vital for improving the functionality of their products. "Tell me what you read and I tell you who you are." – Pierre de La Gorce Voice assistants, such as Alexa, Siri, or Google Assistant, are becoming increasingly popular. Some users are, however, worried about their vocal interactions with these devices being stored in the cloud, together with a textual transcript of every spoken word. But is there an actual threat associated with the collection of these data? And if so, could such threats be prevented?
- Information Technology > Artificial Intelligence > Machine Learning (0.95)
- Information Technology > Artificial Intelligence > Speech > Speech Recognition (0.83)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Personal Assistant Systems (0.54)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.54)
Learning from Dialogue after Deployment: Feed Yourself, Chatbot!
Hancock, Braden, Bordes, Antoine, Mazare, Pierre-Emmanuel, Weston, Jason
The majority of conversations a dialogue agent sees over its lifetime occur after it has already been trained and deployed, leaving a vast store of potential training signal untapped. In this work, we propose the self-feeding chatbot, a dialogue agent with the ability to extract new training examples from the conversations it participates in. As our agent engages in conversation, it also estimates user satisfaction in its responses. When the conversation appears to be going well, the user's responses become new training examples to imitate. When the agent believes it has made a mistake, it asks for feedback; learning to predict the feedback that will be given improves the chatbot's dialogue abilities further. On the PersonaChat chit-chat dataset with over 131k training examples, we find that learning from dialogue with a self-feeding chatbot significantly improves performance, regardless of the amount of traditional supervision.
- Europe > France (0.04)
- North America > Aruba (0.04)
- North America > United States > Texas (0.04)
- (7 more...)
- Personal > Interview (0.69)
- Research Report > Experimental Study (0.67)
- Research Report > New Finding (0.46)
- Leisure & Entertainment (1.00)
- Media (0.93)